Serveur sur les données et bibliothèques médicales au Maghreb (version finale)

Attention, ce site est en cours de développement !
Attention, site généré par des moyens informatiques à partir de corpus bruts.
Les informations ne sont donc pas validées.

Recognizing gestures by learning local motion signatures of HOG descriptors.

Identifieur interne : 000172 ( Main/Exploration ); précédent : 000171; suivant : 000173

Recognizing gestures by learning local motion signatures of HOG descriptors.

Auteurs : Mohamed-Bécha Kaâniche [Tunisie] ; François Brémond

Source :

RBID : pubmed:22997128

Descripteurs français

English descriptors

Abstract

We introduce a new gesture recognition framework based on learning local motion signatures (LMSs) of HOG descriptors introduced by [1]. Our main contribution is to propose a new probabilistic learning-classification scheme based on a reliable tracking of local features. After the generation of these LMSs computed on one individual by tracking Histograms of Oriented Gradient (HOG) [2] descriptor, we learn a codebook of video-words (i.e., clusters of LMSs) using k-means algorithm on a learning gesture video database. Then, the video-words are compacted to a code-book of codewords by the Maximization of Mutual Information (MMI) algorithm. At the final step, we compare the LMSs generated for a new gesture w.r.t. the learned code-book via the k-nearest neighbors (k-NN) algorithm and a novel voting strategy. Our main contribution is the handling of the N to N mapping between codewords and gesture labels within the proposed voting strategy. Experiments have been carried out on two public gesture databases: KTH [3] and IXMAS [4]. Results show that the proposed method outperforms recent state-of-the-art methods.

DOI: 10.1109/TPAMI.2012.19
PubMed: 22997128


Affiliations:


Links toward previous steps (curation, corpus...)


Le document en format XML

<record>
<TEI>
<teiHeader>
<fileDesc>
<titleStmt>
<title xml:lang="en">Recognizing gestures by learning local motion signatures of HOG descriptors.</title>
<author>
<name sortKey="Kaaniche, Mohamed Becha" sort="Kaaniche, Mohamed Becha" uniqKey="Kaaniche M" first="Mohamed-Bécha" last="Kaâniche">Mohamed-Bécha Kaâniche</name>
<affiliation wicri:level="1">
<nlm:affiliation>Higher School of Communications of Tunis (Sup’Com), University of Carthage, Ariana, Tunisia.</nlm:affiliation>
<country xml:lang="fr">Tunisie</country>
<wicri:regionArea>Higher School of Communications of Tunis (Sup’Com), University of Carthage, Ariana</wicri:regionArea>
<wicri:noRegion>Ariana</wicri:noRegion>
</affiliation>
</author>
<author>
<name sortKey="Bremond, Francois" sort="Bremond, Francois" uniqKey="Bremond F" first="François" last="Brémond">François Brémond</name>
</author>
</titleStmt>
<publicationStmt>
<idno type="wicri:source">PubMed</idno>
<date when="2012">2012</date>
<idno type="RBID">pubmed:22997128</idno>
<idno type="pmid">22997128</idno>
<idno type="doi">10.1109/TPAMI.2012.19</idno>
<idno type="wicri:Area/PubMed/Corpus">000648</idno>
<idno type="wicri:explorRef" wicri:stream="PubMed" wicri:step="Corpus" wicri:corpus="PubMed">000648</idno>
<idno type="wicri:Area/PubMed/Curation">000644</idno>
<idno type="wicri:explorRef" wicri:stream="PubMed" wicri:step="Curation">000644</idno>
<idno type="wicri:Area/PubMed/Checkpoint">000648</idno>
<idno type="wicri:explorRef" wicri:stream="Checkpoint" wicri:step="PubMed">000648</idno>
<idno type="wicri:Area/Main/Merge">000172</idno>
<idno type="wicri:Area/Main/Curation">000172</idno>
<idno type="wicri:Area/Main/Exploration">000172</idno>
</publicationStmt>
<sourceDesc>
<biblStruct>
<analytic>
<title xml:lang="en">Recognizing gestures by learning local motion signatures of HOG descriptors.</title>
<author>
<name sortKey="Kaaniche, Mohamed Becha" sort="Kaaniche, Mohamed Becha" uniqKey="Kaaniche M" first="Mohamed-Bécha" last="Kaâniche">Mohamed-Bécha Kaâniche</name>
<affiliation wicri:level="1">
<nlm:affiliation>Higher School of Communications of Tunis (Sup’Com), University of Carthage, Ariana, Tunisia.</nlm:affiliation>
<country xml:lang="fr">Tunisie</country>
<wicri:regionArea>Higher School of Communications of Tunis (Sup’Com), University of Carthage, Ariana</wicri:regionArea>
<wicri:noRegion>Ariana</wicri:noRegion>
</affiliation>
</author>
<author>
<name sortKey="Bremond, Francois" sort="Bremond, Francois" uniqKey="Bremond F" first="François" last="Brémond">François Brémond</name>
</author>
</analytic>
<series>
<title level="j">IEEE transactions on pattern analysis and machine intelligence</title>
<idno type="eISSN">1939-3539</idno>
<imprint>
<date when="2012" type="published">2012</date>
</imprint>
</series>
</biblStruct>
</sourceDesc>
</fileDesc>
<profileDesc>
<textClass>
<keywords scheme="KwdEn" xml:lang="en">
<term>Algorithms (MeSH)</term>
<term>Artificial Intelligence (MeSH)</term>
<term>Gestures (MeSH)</term>
<term>Humans (MeSH)</term>
<term>Image Enhancement (methods)</term>
<term>Image Interpretation, Computer-Assisted (methods)</term>
<term>Movement (physiology)</term>
<term>Pattern Recognition, Automated (methods)</term>
<term>Reproducibility of Results (MeSH)</term>
<term>Sensitivity and Specificity (MeSH)</term>
<term>Video Recording (methods)</term>
<term>Whole Body Imaging (methods)</term>
</keywords>
<keywords scheme="KwdFr" xml:lang="fr">
<term>Algorithmes (MeSH)</term>
<term>Amélioration d'image (méthodes)</term>
<term>Enregistrement sur magnétoscope (méthodes)</term>
<term>Gestes (MeSH)</term>
<term>Humains (MeSH)</term>
<term>Imagerie du corps entier (méthodes)</term>
<term>Intelligence artificielle (MeSH)</term>
<term>Interprétation d'images assistée par ordinateur (méthodes)</term>
<term>Mouvement (physiologie)</term>
<term>Reconnaissance automatique des formes (méthodes)</term>
<term>Reproductibilité des résultats (MeSH)</term>
<term>Sensibilité et spécificité (MeSH)</term>
</keywords>
<keywords scheme="MESH" qualifier="methods" xml:lang="en">
<term>Image Enhancement</term>
<term>Image Interpretation, Computer-Assisted</term>
<term>Pattern Recognition, Automated</term>
<term>Video Recording</term>
<term>Whole Body Imaging</term>
</keywords>
<keywords scheme="MESH" qualifier="méthodes" xml:lang="fr">
<term>Amélioration d'image</term>
<term>Enregistrement sur magnétoscope</term>
<term>Imagerie du corps entier</term>
<term>Interprétation d'images assistée par ordinateur</term>
<term>Reconnaissance automatique des formes</term>
</keywords>
<keywords scheme="MESH" qualifier="physiologie" xml:lang="fr">
<term>Mouvement</term>
</keywords>
<keywords scheme="MESH" qualifier="physiology" xml:lang="en">
<term>Movement</term>
</keywords>
<keywords scheme="MESH" xml:lang="en">
<term>Algorithms</term>
<term>Artificial Intelligence</term>
<term>Gestures</term>
<term>Humans</term>
<term>Reproducibility of Results</term>
<term>Sensitivity and Specificity</term>
</keywords>
<keywords scheme="MESH" xml:lang="fr">
<term>Algorithmes</term>
<term>Gestes</term>
<term>Humains</term>
<term>Intelligence artificielle</term>
<term>Reproductibilité des résultats</term>
<term>Sensibilité et spécificité</term>
</keywords>
</textClass>
</profileDesc>
</teiHeader>
<front>
<div type="abstract" xml:lang="en">We introduce a new gesture recognition framework based on learning local motion signatures (LMSs) of HOG descriptors introduced by [1]. Our main contribution is to propose a new probabilistic learning-classification scheme based on a reliable tracking of local features. After the generation of these LMSs computed on one individual by tracking Histograms of Oriented Gradient (HOG) [2] descriptor, we learn a codebook of video-words (i.e., clusters of LMSs) using k-means algorithm on a learning gesture video database. Then, the video-words are compacted to a code-book of codewords by the Maximization of Mutual Information (MMI) algorithm. At the final step, we compare the LMSs generated for a new gesture w.r.t. the learned code-book via the k-nearest neighbors (k-NN) algorithm and a novel voting strategy. Our main contribution is the handling of the N to N mapping between codewords and gesture labels within the proposed voting strategy. Experiments have been carried out on two public gesture databases: KTH [3] and IXMAS [4]. Results show that the proposed method outperforms recent state-of-the-art methods.</div>
</front>
</TEI>
<affiliations>
<list>
<country>
<li>Tunisie</li>
</country>
</list>
<tree>
<noCountry>
<name sortKey="Bremond, Francois" sort="Bremond, Francois" uniqKey="Bremond F" first="François" last="Brémond">François Brémond</name>
</noCountry>
<country name="Tunisie">
<noRegion>
<name sortKey="Kaaniche, Mohamed Becha" sort="Kaaniche, Mohamed Becha" uniqKey="Kaaniche M" first="Mohamed-Bécha" last="Kaâniche">Mohamed-Bécha Kaâniche</name>
</noRegion>
</country>
</tree>
</affiliations>
</record>

Pour manipuler ce document sous Unix (Dilib)

EXPLOR_STEP=$WICRI_ROOT/Wicri/Sante/explor/MaghrebDataLibMedV2/Data/Main/Exploration
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 000172 | SxmlIndent | more

Ou

HfdSelect -h $EXPLOR_AREA/Data/Main/Exploration/biblio.hfd -nk 000172 | SxmlIndent | more

Pour mettre un lien sur cette page dans le réseau Wicri

{{Explor lien
   |wiki=    Wicri/Sante
   |area=    MaghrebDataLibMedV2
   |flux=    Main
   |étape=   Exploration
   |type=    RBID
   |clé=     pubmed:22997128
   |texte=   Recognizing gestures by learning local motion signatures of HOG descriptors.
}}

Pour générer des pages wiki

HfdIndexSelect -h $EXPLOR_AREA/Data/Main/Exploration/RBID.i   -Sk "pubmed:22997128" \
       | HfdSelect -Kh $EXPLOR_AREA/Data/Main/Exploration/biblio.hfd   \
       | NlmPubMed2Wicri -a MaghrebDataLibMedV2 

Wicri

This area was generated with Dilib version V0.6.38.
Data generation: Wed Jun 30 18:27:05 2021. Site generation: Wed Jun 30 18:34:21 2021